Lightweight Deep Learning Framework for Speech Emotion Recognition

نویسندگان

چکیده

Speech Emotion Recognition (SER) system, which analyzes human utterances to determine a speaker’s emotion, has growing impact on how people and machines interact. Recent growth in human-computer interaction computational intelligence drawn the attention of many researchers Artificial Intelligence (AI) deep learning because its wider applicability several fields, including computer vision, natural language processing, affective computing, among others. Deep models do not need any form manually created features they can automatically extract prospective from input data. models, however, call for lot resources, high processing power, hyper-parameter tuning, making them unsuitable lightweight devices. In this study, we focused developing an efficient model speech emotion recognition with optimized parameters without compromising performance. Our proposed integrates Random Forest Multi-layer Perceptron(MLP) classifiers into VGGNet framework recognition. The was evaluated against other based methods (InceptionV3, ResNet, MobileNetV2, DenseNet) it yielded low complexity optimum experiment carried out three datasets TESS, EMODB, RAVDESS, Mel Frequency Cepstral Coefficient(MFCC) were extracted 6-8 variants emotions namely, Sad, Angry, Happy, Surprise, Neutral, Disgust, Fear, Calm. demonstrated performance 100%, 96%, 86.25% accuracy RAVDESS respectively. This revealed that achieved higher compared recent state-of-the-art found literature.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Speech Emotion Recognition Using Scalogram Based Deep Structure

Speech Emotion Recognition (SER) is an important part of speech-based Human-Computer Interface (HCI) applications. Previous SER methods rely on the extraction of features and training an appropriate classifier. However, most of those features can be affected by emotionally irrelevant factors such as gender, speaking styles and environment. Here, an SER method has been proposed based on a concat...

متن کامل

Efficient Emotion Recognition from Speech Using Deep Learning on Spectrograms

We present a new implementation of emotion recognition from the para-lingual information in the speech, based on a deep neural network, applied directly to spectrograms. This new method achieves higher recognition accuracy compared to previously published results, while also limiting the latency. It processes the speech input in smaller segments – up to 3 seconds, and splits a longer input into...

متن کامل

Active learning for dimensional speech emotion recognition

State-of-the-art dimensional speech emotion recognition systems are trained using continuously labelled instances. The data labelling process is labour intensive and time-consuming. In this paper, we propose to apply active learning to reduce according efforts: The unlabelled instances are evaluated automatically, and only the most informative ones are intelligently picked by an informativeness...

متن کامل

Feature Transfer Learning for Speech Emotion Recognition

Speech Emotion Recognition (SER) has achieved some substantial progress in the past few decades since the dawn of emotion and speech research. In many aspects, various research efforts have been made in an attempt to achieve human-like emotion recognition performance in real-life settings. However, with the availability of speech data obtained from different devices and varied acquisition condi...

متن کامل

Representation Learning for Speech Emotion Recognition

Speech emotion recognition is an important problem with applications as varied as human-computer interfaces and affective computing. Previous approaches to emotion recognition have mostly focused on extraction of carefully engineered features and have trained simple classifiers for the emotion task. There has been limited effort at representation learning for affect recognition, where features ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2023

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2023.3297269